visual analytic system
SigTime: Learning and Visually Explaining Time Series Signatures
Huang, Yu-Chia, Chen, Juntong, Liu, Dongyu, Ma, Kwan-Liu
Understanding and distinguishing temporal patterns in time series data is essential for scientific discovery and decision-making. For example, in biomedical research, uncovering meaningful patterns in physiological signals can improve diagnosis, risk assessment, and patient outcomes. However, existing methods for time series pattern discovery face major challenges, including high computational complexity, limited interpretability, and difficulty in capturing meaningful temporal structures. To address these gaps, we introduce a novel learning framework that jointly trains two Transformer models using complementary time series representations: shapelet-based representations to capture localized temporal structures and traditional feature engineering to encode statistical properties. The learned shapelets serve as interpretable signatures that differentiate time series across classification labels. Additionally, we develop a visual analytics system -- SigTIme -- with coordinated views to facilitate exploration of time series signatures from multiple perspectives, aiding in useful insights generation. We quantitatively evaluate our learning framework on eight publicly available datasets and one proprietary clinical dataset. Additionally, we demonstrate the effectiveness of our system through two usage scenarios along with the domain experts: one involving public ECG data and the other focused on preterm labor analysis.
- North America > United States > California > Yolo County > Davis (0.14)
- Asia > Taiwan (0.04)
- South America > Paraguay > Asunción > Asunción (0.04)
- (3 more...)
- Overview (0.92)
- Research Report > Experimental Study (0.67)
- Research Report > New Finding (0.46)
AuraGenome: An LLM-Powered Framework for On-the-Fly Reusable and Scalable Circular Genome Visualizations
Zhang, Chi, Dong, Yu, Wang, Yang, Han, Yuetong, Shan, Guihua, Tang, Bixia
Circular genome visualizations are essential for exploring structural variants and gene regulation. However, existing tools often require complex scripting and manual configuration, making the process time-consuming, error-prone, and difficult to learn. To address these challenges, we introduce AuraGenome, an LLM-powered framework for rapid, reusable, and scalable generation of multi-layered circular genome visualizations. AuraGenome combines a semantic-driven multi-agent workflow with an interactive visual analytics system. The workflow employs seven specialized LLM-driven agents, each assigned distinct roles such as intent recognition, layout planning, and code generation, to transform raw genomic data into tailored visualizations. The system supports multiple coordinated views tailored for genomic data, offering ring, radial, and chord-based layouts to represent multi-layered circular genome visualizations. In addition to enabling interactions and configuration reuse, the system supports real-time refinement and high-quality report export. We validate its effectiveness through two case studies and a comprehensive user study. AuraGenome is available at: https://github.com/Darius18/AuraGenome.
- Workflow (1.00)
- Questionnaire & Opinion Survey (0.87)
- Research Report > Experimental Study (0.68)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Pharmaceuticals & Biotechnology (1.00)
VISLIX: An XAI Framework for Validating Vision Models with Slice Discovery and Analysis
Yan, Xinyuan, Xuan, Xiwei, Ono, Jorge Piazentin, Guo, Jiajing, Mohanty, Vikram, Kumar, Shekar Arvind, Gou, Liang, Wang, Bei, Ren, Liu
Real-world machine learning models require rigorous evaluation before deployment, especially in safety-critical domains like autonomous driving and surveillance. The evaluation of machine learning models often focuses on data slices, which are subsets of the data that share a set of characteristics. Data slice finding automatically identifies conditions or data subgroups where models underperform, aiding developers in mitigating performance issues. Despite its popularity and effectiveness, data slicing for vision model validation faces several challenges. First, data slicing often needs additional image metadata or visual concepts, and falls short in certain computer vision tasks, such as object detection. Second, understanding data slices is a labor-intensive and mentally demanding process that heavily relies on the expert's domain knowledge. Third, data slicing lacks a human-in-the-loop solution that allows experts to form hypothesis and test them interactively. To overcome these limitations and better support the machine learning operations lifecycle, we introduce VISLIX, a novel visual analytics framework that employs state-of-the-art foundation models to help domain experts analyze slices in computer vision models. Our approach does not require image metadata or visual concepts, automatically generates natural language insights, and allows users to test data slice hypothesis interactively. We evaluate VISLIX with an expert study and three use cases, that demonstrate the effectiveness of our tool in providing comprehensive insights for validating object detection models.
- North America > United States > Utah (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > United States > California > Yolo County > Davis (0.04)
- (4 more...)
- Research Report (1.00)
- Overview (0.67)
Trust Calibration as a Function of the Evolution of Uncertainty in Knowledge Generation: A Survey
User trust is a crucial consideration in designing robust visual analytics systems that can guide users to reasonably sound conclusions despite inevitable biases and other uncertainties introduced by the human, the machine, and the data sources which paint the canvas upon which knowledge emerges. A multitude of factors emerge upon studied consideration which introduce considerable complexity and exacerbate our understanding of how trust relationships evolve in visual analytics systems, much as they do in intelligent sociotechnical systems. A visual analytics system, however, does not by its nature provoke exactly the same phenomena as its simpler cousins, nor are the phenomena necessarily of the same exact kind. Regardless, both application domains present the same root causes from which the need for trustworthiness arises: Uncertainty and the assumption of risk. In addition, visual analytics systems, even more than the intelligent systems which (traditionally) tend to be closed to direct human input and direction during processing, are influenced by a multitude of cognitive biases that further exacerbate an accounting of the uncertainties that may afflict the user's confidence, and ultimately trust in the system. In this article we argue that accounting for the propagation of uncertainty from data sources all the way through extraction of information and hypothesis testing is necessary to understand how user trust in a visual analytics system evolves over its lifecycle, and that the analyst's selection of visualization parameters affords us a simple means to capture the interactions between uncertainty and cognitive bias as a function of the attributes of the search tasks the analyst executes while evaluating explanations. We sample a broad cross-section of the literature from visual analytics, human cognitive theory, and uncertainty, and attempt to synthesize a useful perspective.
- North America > United States > Illinois (0.04)
- North America > United States > Washington > Benton County > Richland (0.04)
- Information Technology > Human Computer Interaction > Interfaces (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Scientific Discovery (0.34)
Apple ML Researchers Develop 'Neo': A Visual Analytics System That Enables Machine Learning Practitioners To Generalize Confusion Matrix Visualization to Hierarchical and Multi-Output Labels
In Machine Learning (ML), model evaluation is the most challenging step. The confusion matrix is one of the globally utilized performance metrics to evaluate the model for classification tasks. It is also a visualization tool that many ML courses and researchers have used. Moreover, it is a table with two dimensions, i.e., actual class label and predicted class label. The actual class label is represented by a row, while a column in the confusion matrix represents the predicted class label.
Data Science Engineer (Decision Science Visualization team - Remote)
As a Data Science Engineer in our Decision Sciences Visualization team, you will create innovative visual analytics systems that reveal, explore and explain complex patterns and phenomena from Epsilon's peta-scale and massively-dimensional digital marketing ecosystem. You will reduce this complexity into sophisticated, interactive visual metaphors and stories that demonstrate the business value of Epsilon's platform directly to our stakeholders and global customers (which includes some of the world's largest brands). Our visualization challenges span hundreds of millions of highly detailed individual profiles across hundreds of millions of web sites and apps tied together by sophisticated real-time analytics that make near instantaneous decisions on messaging to those profiles trillions of times a day. This provides an incredibly rich environment of business questions and answers that are hidden within petabytes of data. In this role, you will be a key member of a multi-disciplinary R&D team creating a full-stack visual analytics system that connects to big data platforms and machine learning systems to present complex enterprise-critical information in an engaging, intuitive and informative manner.
- Information Technology > Data Science > Data Mining > Big Data (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
A Visual Analytics System for Multi-model Comparison on Clinical Data Predictions
Li, Yiran, Fujiwara, Takanori, Choi, Yong K., Kim, Katherine K., Ma, Kwan-Liu
There is a growing trend of applying machine learning methods to medical datasets in order to predict patients' future status. Although some of these methods achieve high performance, challenges still exist in comparing and evaluating different models through their interpretable information. Such analytics can help clinicians improve evidence-based medical decision making. In this work, we develop a visual analytics system that compares multiple models' prediction criteria and evaluates their consistency. Using our system, knowledge can be generated on how differently each model made the predictions and how confidently we can rely on each model's prediction for a certain patient. Through a case study of a publicly available clinical dataset, we demonstrate the effectiveness of our visual analytics system to assist clinicians and researchers in comparing and quantitatively evaluating different machine learning methods.
- North America > United States > California > Yolo County > Davis (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
Comparative Visual Analytics for Assessing Medical Records with Sequence Embedding
Guo, Rongchen, Fujiwara, Takanori, Li, Yiran, Lima, Kelly M., Sen, Soman, Tran, Nam K., Ma, Kwan-Liu
Machine learning for data-driven diagnosis has been actively studied in medicine to provide better healthcare. Supporting analysis of a patient cohort similar to a patient under treatment is a key task for clinicians to make decisions with high confidence. However, such analysis is not straightforward due to the characteristics of medical records: high dimensionality, irregularity in time, and sparsity. To address this challenge, we introduce a method for similarity calculation of medical records. Our method employs event and sequence embeddings. While we use an autoencoder for the event embedding, we apply its variant with the self-attention mechanism for the sequence embedding. Moreover, in order to better handle the irregularity of data, we enhance the self-attention mechanism with consideration of different time intervals. We have developed a visual analytics system to support comparative studies of patient records. To make a comparison of sequences with different lengths easier, our system incorporates a sequence alignment method. Through its interactive interface, the user can quickly identify patients of interest and conveniently review both the temporal and multivariate aspects of the patient records. We demonstrate the effectiveness of our design and system with case studies using a real-world dataset from the neonatal intensive care unit of UC Davis.
AI System that predicts traffic conditions
UNIST scientists have recently developed an interactive visual analytics system that enables traffic congestion exploration, surveillance, and forecasting based on vehicle detector data. Their system can predict traffic conditions for the next 5 to 15 minutes at an error rate of fewer than four kilometers an hour. This intelligent visual analytics system empowers traffic congestion exploration, observation, and determining dependent on vehicle detector information. Through domain expert collaboration, we have extricated task requirements, consolidated the Long-Short Term Memory (LSTM) model for congestion forecasting, and designed a weighting technique for distinguishing the reasons for congestion and congestion propagation directions. The system then visualized the traffic situation for easier comprehension: Congestion levels and average driving speed, for instance, are described using colors and shapes.
- Transportation (0.81)
- Consumer Products & Services > Travel (0.69)
Visual Analytics for Explainable Deep Learning
Jaegul Choo Korea University Shixia Liu Tsinghua University Recently, deep learning has been advancing the state of the art in artificial intelligence to a new level, and humans rely on artificial intelligence techniques more than ever. However, even with such unprecedented advancements, the lack of explanation regarding the decisions made by deep learning models and absence of control over their internal processes act as major drawbacks in critical decision-making processes, such as precision medicine and law enforcement. In response, efforts are being made to make deep learning interpretable and controllable by humans. In this paper, we review visual analytics, information visualization, and machine learning perspectives relevant to this aim, and discuss potential challenges and future research directions. Deep learning has had a considerable impact on various long-running artificial intelligence problems, including computer vision, speech recognition and synthesis, and natural language understanding and generation [1].
- Europe (0.14)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Asia > China (0.04)
- Overview (0.66)
- Research Report (0.50)